58 research outputs found

    Rhythmic variability between speakers:articulatory, prosodic and linguistic factors

    Get PDF
    Between-speaker variability of acoustically measurable speech rhythm [%V, ΔV(ln), ΔC(ln), and Δpeak(ln)] was investigated when within-speaker variability of (a) articulation rate and (b) linguistic structural characteristics was introduced. To study (a), 12 speakers of Standard German read seven lexically identical sentences under five different intended tempo conditions (very slow, slow, normal, fast, very fast). To study (b), 16 speakers of Zurich Swiss German produced 16 spontaneous utterances each (256 in total) for which transcripts were made and then read by all speakers (4096 sentences; 16 speaker × 256 sentences). Between-speaker variability was tested using analysis of variance with repeated measures on within-speaker factors. Results revealed strong and consistent between-speaker variability while within-speaker variability as a function of articulation rate and linguistic characteristics was typically not significant. It was concluded that between-speaker variability of acoustically measurable speech rhythm is strong and robust against various sources of within-speaker variability. Idiosyncratic articulatory movements were found to be the most plausible factor explaining between-speaker differences

    Vowel recognition at fundamental frequencies up to 1 kHz reveals point vowels as acoustic landmarks

    Get PDF
    The phonological function of vowels can be maintained at fundamental frequencies (fo) up to 880 Hz [Friedrichs, Maurer, and Dellwo (2015). J. Acoust. Soc. Am. 138, EL36–EL42]. Here, the influence of talker variability and multiple response options on vowel recognition at high fos is assessed. The stimuli (n = 264) consisted of eight isolated vowels (/i y e ø ε a o u/) produced by three female native German talkers at 11 fos within a range of 220–1046 Hz. In a closed-set identification task, 21 listeners were presented excised 700-ms vowel nuclei with quasi-flat fo contours and resonance trajectories. The results show that listeners can identify the point vowels /i a u/ at fos up to almost 1 kHz, with a significant decrease for the vowels /y ε/ and a drop to chance level for the vowels /e ø o/ toward the upper fos. Auditory excitation patterns reveal highly differentiable representations for /i a u/ that can be used as landmarks for vowel category perception at high fos. These results suggest that theories of vowel perception based on overall spectral shape will provide a fuller account of vowel perception than those based solely on formant frequency patterns

    Rhythm and Vowel Quality in Accents of English

    Get PDF
    In a sample of 27 speakers of Scottish Standard English two notoriously variable consonantal features are investigated: the contrast of /m/ and /w/ and non-prevocalic /r/, the latter both in terms of its presence or absence and the phonetic form it takes, if present. The pattern of realisation of non-prevocalic /r/ largely confirms previously reported findings. But there are a number of surprising results regarding the merger of /m/ and /w/ and the loss of non-prevocalic /r/: While the former is more likely to happen in younger speakers and females, the latter seems more likely in older speakers and males. This is suggestive of change in progress leading to a loss of the /m/ - /w/ contrast, while the variation found in non-prevocalic /r/ follows an almost inverse sociolinguistic pattern that does not suggest any such change and is additionally largely explicable in language-internal terms. One phenomenon requiring further investigation is the curious effect direct contact with Southern English accents seems to have on non-prevocalic /r/: innovation on the structural level (i.e. loss) and conservatism on the realisational level (i.e. increased incidence of [r] and [r]) appear to be conditioned by the same sociolinguistic factors

    Speech rhythm: a metaphor?

    Get PDF
    Is speech rhythmic? In the absence of evidence for a traditional view that languages strive to coordinate either syllables or stress-feet with regular time intervals, we consider the alternative that languages exhibit contrastive rhythm subsisting merely in the alternation of stronger and weaker elements. This is initially plausible, particularly for languages with a steep ‘prominence gradient’, i.e. a large disparity between stronger and weaker elements; but we point out that alternation is poorly achieved even by a ‘stress-timed’ language such as English, and, historically, languages have conspicuously failed to adopt simple phonological remedies that would ensure alternation. Languages seem more concerned to allow ‘syntagmatic contrast’ between successive units and to use durational effects to support linguistic functions than to facilitate rhythm. Furthermore, some languages (e.g. Tamil, Korean) lack the lexical prominence which would most straightforwardly underpin prominence alternation. We conclude that speech is not incontestibly rhythmic, and may even be antirhythmic. However, its linguistic structure and patterning allow the metaphorical extension of rhythm in varying degrees and in different ways depending on the language, and that it is this analogical process which allows speech to be matched to external rhythms

    The Zurich Corpus of Vowel and Voice Quality, Version 1.0

    Get PDF
    Existing databases of isolated vowel sounds or vowel sounds embedded in consonantal context generally document only limited variation of basic production parameters. Thus, concerning the possible variation range of vowel and voice quality-related sound characteristics, there is a lack of broad phenomenological and descriptive references that allow for a comprehensive understanding of vowel acoustics and for an evaluation of the extent to which corresponding existing approaches and models can be generalised. In order to contribute to the building up of such references, a novel database of vowel sounds that exceeds any existing collection by size and diversity of vocalic characteristics is presented here, comprised of c. 34 600 utterances of 70 speakers (46 nonprofessional speakers, children, women and men, and 24 professional actors/actresses and singers of straight theatre, contemporary singing, and European classical singing). The database focuses on sounds of the long Standard German vowels /i-y-e-ø-a-o-u/ produced with varying basic production parameters such as phonation type, vocal effort, fundamental frequency, vowel context and speaking or singing style. In addition, a read text and, for professionals, songs are also included. The database is accessible for scientific use, and further extensions are in progress

    The effects of delayed auditory and visual feedback on speech production

    Get PDF
    Monitoring the sensory consequences of articulatory movements supports speaking. For example, delaying auditory feedback of a speaker's voice disrupts speech production. Also, there is evidence that this disruption may be decreased by immediate visual feedback, i.e., seeing one's own articulatory movements. It is, however, unknown whether delayed visual feedback affects speech production in fluent speakers. Here, the effects of delayed auditory and visual feedback on speech fluency (i.e., speech rate and errors), vocal control (i.e., intensity and pitch), and speech rhythm were investigated. Participants received delayed (by 200 ms) or immediate auditory feedback, while repeating sentences. Moreover, they received either no visual feedback, immediate visual feedback, or delayed visual feedback (by 200, 400, and 600 ms). Delayed auditory feedback affected fluency, vocal control, and rhythm. Immediate visual feedback had no effect on any of the speech measures when it was combined with delayed auditory feedback. Delayed visual feedback did, however, affect speech fluency when it was combined with delayed auditory feedback. In sum, the findings show that delayed auditory feedback disrupts fluency, vocal control, and rhythm and that delayed visual feedback can strengthen the disruptive effect of delayed auditory feedback on fluency

    Electron affinity of Li: A state-selective measurement

    Get PDF
    We have investigated the threshold of photodetachment of Li^- leading to the formation of the residual Li atom in the 2p2P2p ^2P state. The excited residual atom was selectively photoionized via an intermediate Rydberg state and the resulting Li^+ ion was detected. A collinear laser-ion beam geometry enabled both high resolution and sensitivity to be attained. We have demonstrated the potential of this state selective photodetachment spectroscopic method by improving the accuracy of Li electron affinity measurements an order of magnitude. From a fit to the Wigner law in the threshold region, we obtained a Li electron affinity of 0.618 049(20) eV.Comment: 5 pages,6 figures,22 reference

    The Role of Native Language and the Fundamental Design of the Auditory System in Detecting Rhythm Changes

    Get PDF
    Accepted December 13, 2018Purpose: We investigated whether rhythm discrimination is mainly driven by the native language of the listener or by the fundamental design of the human auditory system and universal cognitive mechanisms shared by all people irrespective of rhythmic patterns in their native language. Method: In multiple experiments, we asked participants to listen to 2 continuous acoustic sequences and to determine whether their rhythms were the same or different (AX discrimination). Participants were native speakers of 4 languages with different rhythmic properties (Spanish, French, English, and German) to understand whether the predominant rhythmic patterns of a native language affect sensitivity, bias, and reaction time in detecting rhythmic changes in linguistic (Experiment 2) and in nonlinguistic (Experiments 1 and 2) acoustic sequences. We examined sensitivity and bias measures, as well as reaction times. We also computed Bayes factors in order to assess the effect of native language. Results: All listeners performed better (i.e., responded faster and manifested higher sensitivity and accuracy) when detecting the presence or absence of a rhythm change when the 1st stimulus in an AX test pair exhibited regular rhythm (i.e., a syllable-timed rhythmic pattern) than when the 1st stimulus exhibited irregular rhythm (i.e., stress-timed rhythmic pattern). This result pattern was observed both on linguistic and nonlinguistic stimuli and was not modulated by the native language of the participant. Conclusion: We conclude that rhythm change detection is a fundamental function of a processing system that relies on general auditory mechanisms and is not modulated by linguistic experience.The authors acknowledge support from Spanish Ministry of Economy and Competitiveness Grant PSI2017-82563-P (awarded to A. G. S.), the “Severo Ochoa” Programme for Centres/Units of Excellence in R&D Grant SEV-2015-490 (BCBL), and the Basque Foundation for Science Grant IKERBASQUE (awarded to A. G. S. and M. O.). D. M. G. was supported by Grant PIA/Basal FB0003 from the Chilean Research Council. L. P. was supported by the Spanish Ministry of Economy and Competitiveness via Juan de la Cierva fellowship

    Assessment of rhythm

    No full text
    corecore